具有波束成型的天线阵列在较高的载波频率下克服了高空间路径损耗。但是,必须正确对齐光束,以确保用户设备(UE)辐射(并接收)最高功率。尽管有一些方法可以通过某种形式的层次搜索来详尽地搜索最佳光束,但它们可能很容易返回具有小型梁增益的本地最佳解决方案。其他方法通过利用上下文信息(例如UE的位置或来自相邻基站(BS)的信息的位置)来解决此问题,但是计算和传达此附加信息的负担可能很高。迄今为止,基于机器学习的方法受到随附的培训,性能监控和部署复杂性的影响,从而阻碍了其规模的应用。本文提出了一种解决初始光束发现问题的新方法。它是可扩展的,易于调整和实施。我们的算法基于一个推荐系统,该系统基于培训数据集将组(即UES)和偏好(即来自代码簿中的光束)关联。每当需要提供新的UE时,我们的算法都会返回此用户群集中的最佳光束。我们的仿真结果证明了我们方法的效率和鲁棒性,不仅在单个BS设置中,而且在需要几个BS之间协调的设置中。我们的方法在给定任务中始终优于标准基线算法。
translated by 谷歌翻译
The deep learning community has witnessed an exponentially growing interest in self-supervised learning (SSL). However, it still remains unexplored how to build a framework for learning useful representations of raw music waveforms in a self-supervised manner. In this work, we design Music2Vec, a framework exploring different SSL algorithmic components and tricks for music audio recordings. Our model achieves comparable results to the state-of-the-art (SOTA) music SSL model Jukebox, despite being significantly smaller with less than 2% of parameters of the latter. The model will be released on Huggingface(Please refer to: https://huggingface.co/m-a-p/music2vec-v1)
translated by 谷歌翻译
Human-technology collaboration relies on verbal and non-verbal communication. Machines must be able to detect and understand the movements of humans to facilitate non-verbal communication. In this article, we introduce ongoing research on human activity recognition in intralogistics, and show how it can be applied in industrial settings. We show how semantic attributes can be used to describe human activities flexibly and how context informantion increases the performance of classifiers to recognise them automatically. Beyond that, we present a concept based on a cyber-physical twin that can reduce the effort and time necessary to create a training dataset for human activity recognition. In the future, it will be possible to train a classifier solely with realistic simulation data, while maintaining or even increasing the classification performance.
translated by 谷歌翻译
The problem of predicting driver attention from the driving perspective is gaining the increasing research focuses due to its remarkable significance for autonomous driving and assisted driving systems. Driving experience is extremely important for driver attention prediction, a skilled driver is able to effortlessly predict oncoming danger (before it becomes salient) based on driving experience and quickly pay attention on the corresponding zones. However, the nonobjective driving experience is difficult to model, so a mechanism simulating driver experience accumulation procedure is absent in existing methods, and the existing methods usually follow the technique line of saliency prediction methods to predict driver attention. In this paper, we propose a FeedBack Loop Network (FBLNet), which attempts to model the driving experience accumulation procedure. By over-and-over iterations, FBLNet generates the incremental knowledge that carries rich historically-accumulative long-term temporal information. The incremental knowledge to our model is like the driving experience to humans. Under the guidance of the incremental knowledge, our model fuses the CNN feature and Transformer feature that are extracted from the input image to predict driver attention. Our model exhibits solid advantage over existing methods, achieving an average 10.3% performance improvement on three public datasets.
translated by 谷歌翻译
This paper shows the implementation of reinforcement learning (RL) in commercial flowsheet simulator software (Aspen Plus V12) for designing and optimising a distillation sequence. The aim of the SAC agent was to separate a hydrocarbon mixture in its individual components by utilising distillation. While doing so it tries to maximise the profit produced by the distillation sequence. All actions of the agent were set by the SAC agent in Python and communicated in Aspen Plus via an API. Here the distillation column was simulated by use of the build-in RADFRAC column. With this a connection was established for data transfer between Python and Aspen and the agent succeeded to show learning behaviour, while increasing profit. Although results were generated, the use of Aspen was slow (190 hours) and Aspen was found unsuitable for parallelisation. This makes that Aspen is incompatible for solving RL problems. Code and thesis are available at https://github.com/lollcat/Aspen-RL
translated by 谷歌翻译
由于自动驾驶,物联网和流媒体服务的快速发展,现代通信系统必须应对各种渠道条件以及用户和设备的稳步增加。这以及仍在上升的带宽需求只能通过智能网络自动化来满足,这需要高度灵活和盲目的收发器算法。为了应对这些挑战,我们提出了一种新颖的自适应均衡计划,该计划通过训练用对抗性网络训练均衡器来利用深度学习的繁荣进步。该学习仅基于发射信号的统计数据,因此它对通道模型的实际发送符号和不可知论是盲目的。所提出的方法独立于均衡器拓扑,并实现了强大的基于神经网络的均衡器的应用。在这项工作中,我们证明了这一概念在对线性和非线性传输通道的模拟中,并证明了拟议的盲目学习方案的能力,可以接近非盲均衡器的性能。此外,我们提供了理论观点,并强调了方法的挑战。
translated by 谷歌翻译
在这项工作的过程中,我们检查了塑料轮廓挤出的过程,其中聚合物熔体在所谓的挤出模中形状,并通过在下游校准单元中固化为其形状。更精确,我们专注于数据驱动的减少订单模型(ROM),目的是预测校准单元内挤出的轮廓内的温度分布。在其中,ROM是我们基于预测的过程控制总体目标的第一步,以避免最终产品的不想要的扭曲和损坏。
translated by 谷歌翻译
增强学习(RL)是多能管理系统的有前途的最佳控制技术。它不需要先验模型 - 降低了前期和正在进行的项目特定工程工作,并且能够学习基础系统动力学的更好表示。但是,香草RL不能提供约束满意度的保证 - 导致其在安全至关重要的环境中产生各种不安全的互动。在本文中,我们介绍了两种新颖的安全RL方法,即SafeFallback和Afvafe,其中安全约束配方与RL配方脱钩,并且提供了硬构成满意度,可以保证在培训(探索)和开发过程中(近距离) )最佳政策。在模拟的多能系统案例研究中,我们已经表明,这两种方法均与香草RL基准相比(94,6%和82,8%,而35.5%)和香草RL基准相比明显更高的效用(即有用的政策)开始。提出的SafeFallback方法甚至可以胜过香草RL基准(102,9%至100%)。我们得出的结论是,这两种方法都是超越RL的安全限制处理技术,正如随机代理所证明的,同时仍提供坚硬的保证。最后,我们向I.A.提出了基本的未来工作。随着更多数据可用,改善约束功能本身。
translated by 谷歌翻译
当前用于多模式任务的体系结构,例如视觉问题回答的较高复杂性。结果,这些架构很难训练,需要高度的计算资源。为了解决这些问题,我们提出了一个基于夹的体系结构,该体系结构不需要对功能提取器进行任何微调。简单的线性分类器用于图像和文本编码器的串联特征。在训练过程中,添加了辅助损失,该辅助损失可在答案类型上运行。然后将结果分类用作答案类选择的注意门。在Vizwiz 2022视觉问题回答挑战中,我们在任务1上获得了60.15%的准确性:预测任务2:预测视觉问题的可回答性的视觉问题和AP得分为83.78%。
translated by 谷歌翻译
监督的机器学习方法需要在训练阶段最小化损失功能。顺序数据在许多研究领域中无处不在,并且通常通过为表格数据设计的基于欧几里得距离的损失函数处理。对于平滑的振荡数据,这些常规方法缺乏对同时惩罚幅度,频率和相位预测误差的能力,并且倾向于偏向振幅误差。我们将表面相似性参数(SSP)作为一种新型损耗函数引入,对于平滑振荡序列的训练机器学习模型特别有用。我们对混沌时空动力学系统进行的广泛实验表明,SSP有益于塑造梯度,从而加速训练过程,减少最终预测误差,增加重量初始化的鲁棒性以及与使用经典损失功能相比,实施更强的正则化效果。结果表明,新型损失度量的潜力,特别是对于高度复杂和混乱的数据,例如由非线性二维Kuramoto-Sivashinsky方程以及流体中分散表面重力波的线性传播所引起的数据。
translated by 谷歌翻译